The roles and recognition of Haptic-Ostensive actions in collaborative multimodal human-human dialogues

نویسندگان

  • Lin Chen
  • Maria Javaid
  • Barbara Di Eugenio
  • Milos Zefran
چکیده

The RoboHelper project has the goal of developing assistive robots for the elderly. One crucial component of such a robot is a multimodal dialogue architecture, since collaborative task-oriented human-human dialogue is inherently multimodal. In this paper, we focus on a specific type of interaction, Haptic-Ostensive (H-O) actions, that are pervasive in collaborative dialogue. H-O actions manipulate objects, but they also often perform a referring function. We collected 20 collaborative task-oriented human-human dialogues between a helper and an elderly person in a realistic setting. To collect the haptic signals, we developed an unobtrusive sensory glove with pressure sensors. Multiple annotations were then conducted to build the Find corpus. Supervised machine learning was applied to these annotations in order to develop reference resolution and dialogue act classification modules. Both corpus analysis, and these two modules show that H-O actions play a crucial role in interaction: models that include H-O actions, and other extra-linguistic information such as pointing gestures, perform better. For true human-robot interaction, all communicative intentions must of course be recognized in real time, not on the basis of annotated categories. To demonstrate that our corpus analysis is not an end in itself, but can inform actual human-robot interaction, the last part of our paper presents additional experiments on recognizing H-O actions from the haptic signals measured through the sensory glove. We show that even though pressure sensors are relatively imprecise and the data provided by the glove is noisy, the classification algorithms can successfully identify actions of interest within subjects. c © 2015 Published by Elsevier Ltd.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Multimodal Communication Interface for Assistive Robots for the Elderly

The RoboHelper project aims at building an effective and user friendly communication interface for assistive robots that can help the elderly live independently at home. Such communication interface should incorporate multiple modalities of communication, since collaborative task-oriented human-human communication is inherently multimodal. Data was collected from twenty collaborative task-orien...

متن کامل

The RoboHelper Project: From Multimodal Corpus to Embodiment on a Robot

In this position paper, we describe the RoboHelper project, its findings and our vision for its future. The long-term goal of RoboHelper is to develop assistive robots for the elderly. The main thesis of our work is that such robots must crucially be able to participate in multimodal dialogues. Contributions of our work to date include the ELDERLY-AT-HOME corpus that we collected and annotated....

متن کامل

Co-reference via Pointing and Haptics in Multi-Modal Dialogues

This paper describes our ongoing work on resolving third person pronouns and deictic words in a multi-modal corpus. We show that about two thirds of these referring expressions have antecedents that are introduced by pointing gestures or by haptic-ostensive actions (actions that involve manipulating an object). After describing our annotation scheme, we discuss the co-reference models we learn ...

متن کامل

Supporting group work in crisis management: visually mediated human ^GIS ^ human dialogue

Geospatial information is a fundamental component of many crisis management activities. However, current geospatial technologies do not support work by crisis management personnel, most of whom are not technology specialistsöa key impediment is that the technologies require the user to learn the system's language. In addition, geospatial technologies are not `collaboration friendly'öthey impede...

متن کامل

Affect In Tutoring Dialogues

& This paper is about INES, an intelligent, multimodal tutoring environment, and how we build a tutor agent in the environment that tries to be sensitive to the mental state of the student that interacts with it. The environment was primarily designed to help students practice nursing tasks. For example, one of the implemented tasks is to give a virtual patient a subcutaneous injection. The stu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Computer Speech & Language

دوره 34  شماره 

صفحات  -

تاریخ انتشار 2015